Practical task-oriented parallelism for Gaussian elimination in distributed memory
نویسندگان
چکیده
منابع مشابه
Practical Task-Oriented Parallelism for Gaussian Elimination in Distributed Memory
This paper discusses a methodology for easily and efficiently parallelizing sequential algorithms in linear algebra using cost-effective networks of workstations, where the algorithm lends itself to parallelism. A particular target architecture of interest is the academic student laboratory, which typically contains many networked computers that lay idle at night. A case is made for why a task-...
متن کاملAn analysis of data distribution methods for Gaussian elimination in distributed-memory multicomputers
In multicomputers, an appropriate data distribution is crucial for reducing communication overhead and therefore the overall performance. For this reason, data parallel languages provide programmers with primitives, such as BLOCK and CYCLIC that can be used to distribute data across the distributed memory. However, the languages do not aid the programmer as to how the distribution should be per...
متن کاملAdaptive Parallelism in Distributed Shared Memory Environments
Workstations in networks of workstations (NOWs) are sometimes little used, especially in multi-user environments. Employing their compute power for parallel processing when not used otherwise is an attractive venture, if a practical means to do so can be found. In non-dedicated NOW environments, individual machines become available or unavailable as the workstation “owner” goes away or returns,...
متن کاملRun-Time Techniques for Exploiting Irregular Task Parallelism on Distributed Memory Architectures
Automatic scheduling for directed acyclic graphs (DAG) and its applications for coarse-grained irregular problems such as large n-body simulation have been studied in the literature. However solving irregular problems with mixed granularities such as sparse matrix factorization is challenging since it requires eecient run-time support to execute a DAG schedule. In this paper, we investigate run...
متن کاملA Framework for Exploiting Task and Data Parallelism on Distributed Memory Multicomputers
Distributed Memory Multicomputers (DMMs), such as the IBM SP-2, the Intel Paragon, and the Thinking Machines CM-5, offer significant advantages over shared memory multiprocessors in terms of cost and scalability. Unfortunately, the utilization of all the available computational power in these machines involves a tremendous programming effort on the part of users, which creates a need for sophis...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Linear Algebra and its Applications
سال: 1998
ISSN: 0024-3795
DOI: 10.1016/s0024-3795(97)10035-0